A cloudy sky behind a brightly lit Apple logo in a Manhattan store
An Apple Store in Manhattan.
Mike Segar/Reuters

Apple's Craig Federighi, senior vice president of software engineering, said the company's plan to scan iPhone users' photos for child sexual abuse material has been "widely misunderstood."

"We wish that this had come out a little more clearly for everyone, because we feel very positive and strongly about what we're doing, and we can see that it's been widely misunderstood," Federighi said in a video interview with The Wall Street Journal's Joanna Stern published on Friday.

Apple earlier this month announced a feature that would create digital hashes of images as they're uploaded from iPhones to iCloud accounts. Those hashes would be compared to databases of known child sexual abuse material held by anti-abuse organizations, Apple said.

"Before an image is stored in iCloud Photos, an on-device matching process is performed for that image against the known CSAM hashes," the summary said.

Critics of the plan said it was a misstep for a company that's long made privacy a selling point. The Electronic Frontier Foundation (EFF) last Thursday called the hashing-and-matching update a privacy "backdoor," which could be expanded or exploited. Some Apple employees were worried the feature could be exploited by governments, Reuters reported.

Apple SVP Craig Federighi raises his hand as he speaks on stage in front of a black background
Federighi on stage in 2019.
Jeff Chiu/AP Photo

"I think in no way is this a backdoor," Federighi said in the Journal interview. "I don't understand - I really don't understand that characterization."

The CSAM feature was one of two photo-scanning updates the company announced. The other would scan children's incoming iMessage photos for nudity, alerting parents when children under 12 years old viewed pornographic content.

"I do believe the sound-bite that got out early was, 'Oh my god, Apple is scanning my phone for images.' This is not what is happening," Federighi told the Journal via video. "This is about images stored in the cloud."

Eva Galperin, EFF's director of cybersecurity, said via Twitter that the rollout had not been misunderstood by privacy experts.

"I'd like to take this moment to make it clear to poor Craig that no, I don't misunderstand Apple's plans to check photos in iCloud against NCMEC's database of CSAM," Galperin said.

"It's well-meaning but it's also creating a mechanism that Apple will be forced to use for other things," she added.

Read the original article on Business Insider